Efficient kernel-based variable selection with sparsistency
نویسندگان
چکیده
Variable selection is central to high-dimensional data analysis, and various algorithms have been developed. Ideally, a variable algorithm shall be flexible, scalable, with theoretical guarantee, yet most existing cannot attain these properties at the same time. In this article, three-step developed, involving kernel-based estimation of regression function its gradient functions as well hard thresholding. Its key advantage that it assumes no explicit model assumption, admits general predictor effects, allows for scalable computation, attains desirable asymptotic sparsistency. The proposed can adapted any reproducing kernel Hilbert space (RKHS) different functions, extended interaction slight modification. computational cost only linear in dimension, further improved through parallel computing. sparsistency established RKHS under mild conditions, including Gaussian kernels special cases. effectiveness also supported by variety simulated real examples.
منابع مشابه
Scalable kernel-based variable selection with sparsistency
Variable selection is central to high-dimensional data analysis, and various algorithms have been developed. Ideally, a variable selection algorithm shall be flexible, scalable, and with theoretical guarantee, yet most existing algorithms cannot attain these properties at the same time. In this article, a three-step variable selection algorithm is developed, involving kernel-based estimation of...
متن کاملBayesian Approximate Kernel Regression with Variable Selection
Nonlinear kernel regression models are often used in statistics and machine learning due to greater accuracy than linear models. Variable selection for kernel regression models is a challenge partly because, unlike the linear regression setting, there is no clear concept of an effect size for regression coefficients. In this paper, we propose a novel framework that provides an analog of the eff...
متن کاملScalable Bayesian Kernel Models with Variable Selection
Nonlinear kernels are used extensively in regression models in statistics and machine learning since they often improve predictive accuracy. Variable selection is a challenge in the context of kernel based regression models. In linear regression the concept of an effect size for the regression coefficients is very useful for variable selection. In this paper we provide an analog for the effect ...
متن کاملGradient-based kernel method for feature extraction and variable selection
We propose a novel kernel approach to dimension reduction for supervised learning: feature extraction and variable selection; the former constructs a small number of features from predictors, and the latter finds a subset of predictors. First, a method of linear feature extraction is proposed using the gradient of regression function, based on the recent development of the kernel method. In com...
متن کاملEfficient computer experiment-based optimization through variable selection
A computer experiment-based optimization approach employs design of experiments and statistical modeling to represent a complex objective function that can only be evaluated pointwise by running a computer model. In large-scale applications, the number of variables is huge, and direct use of computer experiments would require an exceedingly large experimental design and, consequently, significa...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Statistica Sinica
سال: 2021
ISSN: ['1017-0405', '1996-8507']
DOI: https://doi.org/10.5705/ss.202019.0401